Rockingham County
The lobstermen teaming up with scientists to save endangered whales
In a game of scientific telephone, if you find the food, you find the whales--and sound the alarm. North Atlantic right whales sometimes gather at Jeffrey's Ledge, a 62-mile-long underwater ridge about 25 miles off the coast of Portsmouth, New Hampshire. Breakthroughs, discoveries, and DIY tips sent six days a week. It was a cold and windy week last January, when a group of Maine lobstermen couldn't haul in their traps from Jeffrey's Ledge. The reason why surprised everyone.
- North America > United States > Maine (0.27)
- North America > United States > New Hampshire > Rockingham County > Portsmouth (0.25)
- North America > United States > Oregon (0.05)
- (5 more...)
- Food & Agriculture > Fishing (0.98)
- Law (0.70)
- North America > Canada > Ontario > Toronto (0.14)
- South America > Brazil > São Paulo (0.04)
- North America > United States > New Hampshire > Rockingham County > Portsmouth (0.04)
- (6 more...)
- Energy (0.46)
- Information Technology (0.46)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Planning & Scheduling (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Optimization (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.99)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning > Gradient Descent (0.32)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.15)
- Asia > Japan > Honshū > Kansai > Kyoto Prefecture > Kyoto (0.04)
- North America > United States > New Hampshire > Rockingham County > Portsmouth (0.04)
- (8 more...)
Do LLMs Really Forget? Evaluating Unlearning with Knowledge Correlation and Confidence Awareness
Wei, Rongzhe, Niu, Peizhi, Hsu, Hans Hao-Hsun, Wu, Ruihan, Yin, Haoteng, Ghassemi, Mohsen, Li, Yifan, Potluru, Vamsi K., Chien, Eli, Chaudhuri, Kamalika, Milenkovic, Olgica, Li, Pan
Machine unlearning techniques aim to mitigate unintended memorization in large language models (LLMs). However, existing approaches predominantly focus on the explicit removal of isolated facts, often overlooking latent inferential dependencies and the non-deterministic nature of knowledge within LLMs. Consequently, facts presumed forgotten may persist implicitly through correlated information. To address these challenges, we propose a knowledge unlearning evaluation framework that more accurately captures the implicit structure of real-world knowledge by representing relevant factual contexts as knowledge graphs with associated confidence scores. We further develop an inference-based evaluation protocol leveraging powerful LLMs as judges; these judges reason over the extracted knowledge subgraph to determine unlearning success. Our LLM judges utilize carefully designed prompts and are calibrated against human evaluations to ensure their trustworthiness and stability. Extensive experiments on our newly constructed benchmark demonstrate that our framework provides a more realistic and rigorous assessment of unlearning performance. Moreover, our findings reveal that current evaluation strategies tend to overestimate unlearning effectiveness. Our code is publicly available at https://github.com/Graph-COM/Knowledge_Unlearning.git.
- Europe > France (0.14)
- North America > United States > North Carolina > Iredell County > Mooresville (0.04)
- North America > United States > North Carolina > Mecklenburg County (0.04)
- (10 more...)
- Banking & Finance (0.93)
- Government (0.67)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.15)
- Asia > Japan > Honshū > Kansai > Kyoto Prefecture > Kyoto (0.04)
- North America > United States > New Hampshire > Rockingham County > Portsmouth (0.04)
- (8 more...)
Reflection-Window Decoding: Text Generation with Selective Refinement
Tang, Zeyu, Chen, Zhenhao, Li, Loka, Song, Xiangchen, Deng, Yunlong, Shen, Yifan, Chen, Guangyi, Spirtes, Peter, Zhang, Kun
The autoregressive decoding for text generation in large language models (LLMs), while widely used, is inherently suboptimal due to the lack of a built-in mechanism to perform refinement and/or correction of the generated content. In this paper, we consider optimality in terms of the joint probability over the generated response, when jointly considering all tokens at the same time. We theoretically characterize the potential deviation of the autoregressively generated response from its globally optimal counterpart that is of the same length. Our analysis suggests that we need to be cautious when noticeable uncertainty arises during text generation, which may signal the sub-optimality of the generation history. To address the pitfall of autoregressive decoding for text generation, we propose an approach that incorporates a sliding reflection window and a pausing criterion, such that refinement and generation can be carried out interchangeably as the decoding proceeds. Our selective refinement framework strikes a balance between efficiency and optimality, and our extensive experimental results demonstrate the effectiveness of our approach.
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > United States > New York (0.04)
- North America > United States > New Hampshire > Rockingham County > Portsmouth (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > Canada > Ontario > Toronto (0.14)
- South America > Brazil > São Paulo (0.04)
- North America > United States > New Hampshire > Rockingham County > Portsmouth (0.04)
- (6 more...)
- Energy (0.46)
- Information Technology (0.46)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Planning & Scheduling (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Optimization (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning > Gradient Descent (0.32)
Tutor CoPilot: A Human-AI Approach for Scaling Real-Time Expertise
Wang, Rose E., Ribeiro, Ana T., Robinson, Carly D., Loeb, Susanna, Demszky, Dora
Generative AI, particularly Language Models (LMs), has the potential to transform real-world domains with societal impact, particularly where access to experts is limited. For example, in education, training novice educators with expert guidance is important for effectiveness but expensive, creating significant barriers to improving education quality at scale. This challenge disproportionately harms students from under-served communities, who stand to gain the most from high-quality education. We introduce Tutor CoPilot, a novel Human-AI approach that leverages a model of expert thinking to provide expert-like guidance to tutors as they tutor. This study is the first randomized controlled trial of a Human-AI system in live tutoring, involving 900 tutors and 1,800 K-12 students from historically under-served communities. Following a preregistered analysis plan, we find that students working with tutors that have access to Tutor CoPilot are 4 percentage points (p.p.) more likely to master topics (p<0.01). Notably, students of lower-rated tutors experienced the greatest benefit, improving mastery by 9 p.p. We find that Tutor CoPilot costs only $20 per-tutor annually. We analyze 550,000+ messages using classifiers to identify pedagogical strategies, and find that tutors with access to Tutor CoPilot are more likely to use high-quality strategies to foster student understanding (e.g., asking guiding questions) and less likely to give away the answer to the student. Tutor interviews highlight how Tutor CoPilot's guidance helps tutors to respond to student needs, though they flag issues in Tutor CoPilot, such as generating suggestions that are not grade-level appropriate. Altogether, our study of Tutor CoPilot demonstrates how Human-AI systems can scale expertise in real-world domains, bridge gaps in skills and create a future where high-quality education is accessible to all students.
- North America > United States > Alaska (0.04)
- North America > United States > New Hampshire > Rockingham County > Portsmouth (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- North America > Mexico > Mexico City > Mexico City (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Health & Medicine (1.00)
- Education > Educational Setting > K-12 Education (1.00)
- Education > Assessment & Standards > Student Performance (1.00)
Data-Driven Prediction and Uncertainty Quantification of PWR Crud-Induced Power Shift Using Convolutional Neural Networks
Furlong, Aidan, Alsafadi, Farah, Palmtag, Scott, Godfrey, Andrew, Wu, Xu
The development of Crud-Induced Power Shift (CIPS) is an operational challenge in Pressurized Water Reactors that is due to the development of crud on the fuel rod cladding. The available predictive tools developed previously, usually based on fundamental physics, are computationally expensive and have shown differing degrees of accuracy. This work proposes a completely top-down approach to predict CIPS instances on an assembly level with reactor-specific calibration built-in. Built using artificial neural networks, this work uses a three-dimensional convolutional approach to leverage the image-like layout of the input data. As a classifier, the convolutional neural network model predicts whether a given assembly will experience CIPS as well as the time of occurrence during a given cycle. This surrogate model is both trained and tested using a combination of calculated core model parameters and measured plant data from Unit 1 of the Catawba Nuclear Station. After the evaluation of its performance using various metrics, Monte Carlo dropout is employed for extensive uncertainty quantification of the model predictions. The results indicate that this methodology could be a viable approach in predicting CIPS with an assembly-level resolution across both clean and afflicted cycles, while using limited computational resources.
- North America > United States > California > San Francisco County > San Francisco (0.14)
- South America > Peru > Lima Department > Lima Province > Lima (0.04)
- North America > United States > Oklahoma > Payne County > Stillwater (0.04)
- (7 more...)
- Health & Medicine (1.00)
- Energy > Power Industry > Utilities > Nuclear (1.00)
DR-RAG: Applying Dynamic Document Relevance to Retrieval-Augmented Generation for Question-Answering
Hei, Zijian, Liu, Weiling, Ou, Wenjie, Qiao, Juyi, Jiao, Junming, Song, Guowen, Tian, Ting, Lin, Yi
Retrieval-Augmented Generation (RAG) has recently demonstrated the performance of Large Language Models (LLMs) in the knowledge-intensive tasks such as Question-Answering (QA). RAG expands the query context by incorporating external knowledge bases to enhance the response accuracy. However, it would be inefficient to access LLMs multiple times for each query and unreliable to retrieve all the relevant documents by a single query. We have found that even though there is low relevance between some critical documents and query, it is possible to retrieve the remaining documents by combining parts of the documents with the query. To mine the relevance, a two-stage retrieval framework called Dynamic-Relevant Retrieval-Augmented Generation (DR-RAG) is proposed to improve document retrieval recall and the accuracy of answers while maintaining efficiency. Additionally, a compact classifier is applied to two different selection strategies to determine the contribution of the retrieved documents to answering the query and retrieve the relatively relevant documents. Meanwhile, DR-RAG call the LLMs only once, which significantly improves the efficiency of the experiment. The experimental results on multi-hop QA datasets show that DR-RAG can significantly improve the accuracy of the answers and achieve new progress in QA systems.
- North America > United States > New York (0.04)
- Europe > France (0.04)
- North America > United States > North Carolina (0.04)
- (15 more...)
- Media > Music (1.00)
- Media > Film (1.00)
- Leisure & Entertainment (1.00)
- Government > Regional Government > North America Government > United States Government (0.46)